On Agnostic Learning of Parities, Monomials, and Halfspaces

نویسندگان

  • Vitaly Feldman
  • Parikshit Gopalan
  • Subhash Khot
  • Ashok Kumar Ponnuswami
چکیده

We study the learnability of several fundamental concept classes in the agnostic learning framework of Haussler [Hau92] and Kearns et al. [KSS94]. We show that under the uniform distribution, agnostically learning parities reduces to learning parities with random classification noise, commonly referred to as the noisy parity problem. Together with the parity learning algorithm of Blum et al. [BKW03], this gives the first nontrivial algorithm for agnostic learning of parities. We use similar techniques to reduce learning of two other fundamental concept classes under the uniform distribution to learning of noisy parities. Namely, we show that learning of DNF expressions reduces to learning noisy parities of just logarithmic number of variables and learning of k-juntas reduces to learning noisy parities of k variables. We give essentially optimal hardness results for agnostic learning of monomials over {0, 1}n and halfspaces over Q. We show that for any constant 2 finding a monomial (halfspace) that agrees with an unknown function on 1/2 + 2 fraction of examples is NP-hard even when there exists a monomial (halfspace) that agrees with the unknown function on 1− 2 fraction of examples. This resolves an open question due to Blum and significantly improves on a number of previous hardness results for these problems. We extend these results to 2 = 2− log 1−λ n (2 = 2− √ log n in the case of halfspaces) for any constant λ > 0 under stronger complexity assumptions. Preliminary versions of the results in this work appeared in [Fel06] and [FGKP06]. ∗Work done while the author was at Harvard University supported by grants from the National Science Foundation NSF-CCR0310882, NSF-CCF-0432037, and NSF-CCF-0427129. †Work done while the author was at Georgia Tech. ‡Supported in part by Subhash Khot’s Microsoft New Faculty Fellowship and Raytheon Fellowship, College of Computing, Georgia Tech.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficiency and Computational Limitations of Learning Algorithms

This thesis presents new positive and negative results concerning the learnability of several well-studied function classes in the Probably Approximately Correct (PAC) model of learning. Learning Disjunctive Normal Form (DNF) expressions in the PAC model is widely considered to be the main open problem in Computational Learning Theory. We prove that PAC learning of DNF expressions by an algorit...

متن کامل

Improved Lower Bounds for Learning Intersections of Halfspaces

We prove new lower bounds for learning intersections of halfspaces, one of the most important concept classes in computational learning theory. Our main result is that any statistical-query algorithm for learning the intersection of √ n halfspaces in n dimensions must make 2 √ n) queries. This is the first non-trivial lower bound on the statistical query dimension for this concept class (the pr...

متن کامل

2 Learning Monotone Boolean Functions

Last time we finished the discussion about KM algorithm and its application. We also covered sparse Fourier representations and k-juntas of parities. In the end we started to talk about learning monotone Boolean functions and influence of such functions. Today We will first finish discussion about learning monotone Boolean functions. Then we will also talk about learning k-juntas of halfspaces....

متن کامل

Learning Halfspaces with Malicious Noise

We give new algorithms for learning halfspaces in the challenging malicious noise model, where an adversary may corrupt both the labels and the underlying distribution of examples. Our algorithms can tolerate malicious noise rates exponentially larger than previous work in terms of the dependence on the dimension n, and succeed for the fairly broad class of all isotropic log-concave distributio...

متن کامل

Submodular functions are noise stable

We show that all non-negative submodular functions have high noise-stability. As a consequence, we obtain a polynomial-time learning algorithm for this class with respect to any product distribution on {−1, 1} (for any constant accuracy parameter ). Our algorithm also succeeds in the agnostic setting. Previous work on learning submodular functions required either query access or strong assumpti...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • SIAM J. Comput.

دوره 39  شماره 

صفحات  -

تاریخ انتشار 2009